The Weighted Sparsity Problem: Complexity and Algorithms

نویسندگان

  • S. Thomas McCormick
  • S. Frank Chang
چکیده

Many optimization algorithms involve repeated processing of a fixed set of linear constraints. If we pre-process the constraint matrix A to make it sparser, algebraic operations should become faster. In many applications there is a priori information about the likelihood that each column will appear in a basis, which can be expressed as weights on the columns. This leads to considering the Weighted Sparsity Problem (WSP): Find a row-equivalent constraint matrix with as small a weight of non-zeros as possible. We show that WSP is NP-hard even with a non-degeneracy assumption, and even if restricted to instances with at most three non-zeros per both row and column. WSP is shown to have a polynomial algorithm when the number of non-zeros per either row or column is limited to at most two. This contrasts with previous results that, assuming only non-degeneracy, the unweighted version of WSP does have a polynomial algorithm (this has proven to be practically useful in tests on real data). The polynomial algorithm for WSP with at most two non-zeros per row or column is based on solving one-row problems via minimum cut calculations, together with a sufficient condition for piecing these one-row solutions together into a global solution.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A NOVEL FUZZY-BASED SIMILARITY MEASURE FOR COLLABORATIVE FILTERING TO ALLEVIATE THE SPARSITY PROBLEM

Memory-based collaborative filtering is the most popular approach to build recommender systems. Despite its success in many applications, it still suffers from several major limitations, including data sparsity. Sparse data affect the quality of the user similarity measurement and consequently the quality of the recommender system. In this paper, we propose a novel user similarity measure based...

متن کامل

Search Based Weighted Multi-Bit Flipping Algorithm for High-Performance Low-Complexity Decoding of LDPC Codes

In this paper, two new hybrid algorithms are proposed for decoding Low Density Parity Check (LDPC) codes. Original version of the proposed algorithms named Search Based Weighted Multi Bit Flipping (SWMBF). The main idea of these algorithms is flipping variable multi bits in each iteration, change in which leads to the syndrome vector with least hamming weight. To achieve this, the proposed algo...

متن کامل

Search Based Weighted Multi-Bit Flipping Algorithm for High-Performance Low-Complexity Decoding of LDPC Codes

In this paper, two new hybrid algorithms are proposed for decoding Low Density Parity Check (LDPC) codes. Original version of the proposed algorithms named Search Based Weighted Multi Bit Flipping (SWMBF). The main idea of these algorithms is flipping variable multi bits in each iteration, change in which leads to the syndrome vector with least hamming weight. To achieve this, the proposed algo...

متن کامل

Applications of two new algorithms of cuckoo optimization (CO) and forest optimization (FO) for solving single row facility layout problem (SRFLP)

Nowadays, due to inherent complexity of real optimization problems, it has always been a challenging issue to develop a solution algorithm to these problems. Single row facility layout problem (SRFLP) is a NP-hard problem of arranging a number of rectangular facilities with varying length on one side of a straight line with aim of minimizing the weighted sum of the distance between all facility...

متن کامل

حل مسئله زمان بندی ماشینهای موازی نامرتبط با اهداف کل زودکرد وزنی و کل دیرکرد وزنی با استفاده از الگوریتم جستجوی پراکنده چند هدفه

The parallel machine scheduling problem is an important and difficult problem to be considered in the real-world situations. Traditionally, this problem consists of the scheduling of a set of independent jobs on parallel machines with the aim of minimizing the maximum job completion. In today's manufacturing systems, in which both early and tardy finishing of job processing are undesired, the o...

متن کامل

Block-Based Compressive Sensing Using Soft Thresholding of Adaptive Transform Coefficients

Compressive sampling (CS) is a new technique for simultaneous sampling and compression of signals in which the sampling rate can be very small under certain conditions. Due to the limited number of samples, image reconstruction based on CS samples is a challenging task. Most of the existing CS image reconstruction methods have a high computational complexity as they are applied on the entire im...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • SIAM J. Discrete Math.

دوره 6  شماره 

صفحات  -

تاریخ انتشار 1993